Finite State Markov Decision Processes with Transfer Entropy Costs
نویسندگان
چکیده
We consider a mathematical framework of finite state Markov Decision Processes (MDPs) in which a weighted sum of the classical state-dependent cost and the transfer entropy from the state random process to the control random process is minimized. Physical interpretations of the considered MDPs are provided in the context of networked control systems theory and non-equilibrium thermodynamics. Based on the dynamic programming principle, we derive an optimality condition comprised of a Kolmogorov forward equation and a Bellman backward equation. As the main contribution, we propose an iterative forward-backward computational procedure similar to the Arimoto-Blahut algorithm to synthesize the optimal policy numerically. Convergence of the algorithm is established. The proposed algorithm is applied to an information-constrained navigation problem over a maze, whereby we study how the price of information alters the optimal decision polices qualitatively.
منابع مشابه
Relative Entropy Rate between a Markov Chain and Its Corresponding Hidden Markov Chain
In this paper we study the relative entropy rate between a homogeneous Markov chain and a hidden Markov chain defined by observing the output of a discrete stochastic channel whose input is the finite state space homogeneous stationary Markov chain. For this purpose, we obtain the relative entropy between two finite subsequences of above mentioned chains with the help of the definition of...
متن کاملPermutation Complexity and Coupling Measures in Hidden Markov Models
In [Haruna, T. and Nakajima, K., 2011. Physica D 240, 13701377], the authors introduced the duality between values (words) and orderings (permutations) as a basis to discuss the relationship between information theoretic measures for finite-alphabet stationary stochastic processes and their permutation analogues. It has been used to give a simple proof of the equality between the entropy rate a...
متن کاملOn the Asymptotic Optimality of Finite Approximations to Markov Decision Processes with Borel Spaces
Abstract. Calculating optimal policies is known to be computationally difficult for Markov decision processes with Borel state and action spaces and for partially observed Markov decision processes even with finite state and action spaces. This paper studies finite-state approximations of discrete time Markov decision processes with Borel state and action spaces, for both discounted and average...
متن کاملCapacity of Finite State Markov Channels with General Inputs
We study new formulae based on Lyapunov exponents for entropy, mutual information, and capacity of finite state discrete time Markov channels. We also develop a method for directly computing mutual information and entropy using continuous state space Markov chains. Our methods allow for arbitrary input processes and channel dynamics, provided both have finite memory. We show that the entropy ra...
متن کاملEntropy and Mutual Information for Markov Channels with General Inputs
We study new formulas based on Lyapunov exponents for entropy, mutual information, and capacity of finite state discrete time Markov channels. We also develop a method for directly computing mutual information and entropy using continuous state space Markov chains. Our methods allow for arbitrary input processes and channel dynamics, provided both have finite memory. We show that the entropy ra...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1708.09096 شماره
صفحات -
تاریخ انتشار 2017